Adversarial examples in random neural networks with general activations
نویسندگان
چکیده
A substantial body of empirical work documents the lack robustness in deep learning models to adversarial examples. Recent theoretical proved that examples are ubiquitous two-layers networks with sub-exponential width and ReLU or smooth activations, multi-layer width. We present a result same type, no restriction on for general locally Lipschitz continuous activations. More precisely, given neural network $f(,\cdot,;\mathbf{\theta})$ random weights $\mathbf{\theta}$, feature vector $\mathbf{x}$, we show an example $\mathbf{x}'$ can be found high probability along direction gradient $\nabla\_{\mathbf{x}}f(\mathbf{x};\mathbf{\theta})$. Our proof is based Gaussian conditioning technique. Instead proving $f$ approximately linear neighborhood characterize joint distribution $f(\mathbf{x};\mathbf{\theta})$ $f(\mathbf{x}';\mathbf{\theta})$ $\mathbf{x}' = \mathbf{x}-s(\mathbf{x})\nabla\_{\mathbf{x}}f(\mathbf{x};\mathbf{\theta})$, where $s(\mathbf{x}) \operatorname{sign}(f(\mathbf{x}; \mathbf{\theta})) \cdot s\_d$ some positive step size $s\_d$.
منابع مشابه
Generating Adversarial Examples with Adversarial Networks
Deep neural networks (DNNs) have been found to be vulnerable to adversarial examples resulting from adding small-magnitude perturbations to inputs. Such adversarial examples can mislead DNNs to produce adversary-selected results. Different attack strategies have been proposed to generate adversarial examples, but how to produce them with high perceptual quality and more efficiently requires mor...
متن کاملFeature Squeezing: Detecting Adversarial Examples in Deep Neural Networks
Although deep neural networks (DNNs) have achieved great success in many tasks, they can often be fooled by adversarial examples that are generated by adding small but purposeful distortions to natural examples. Previous studies to defend against adversarial examples mostly focused on refining the DNN models, but have either shown limited success or required expensive computation. We propose a ...
متن کاملManifold Regularized Deep Neural Networks using Adversarial Examples
Learning meaningful representations using deep neural networks involves designing efficient training schemes and well-structured networks. Currently, the method of stochastic gradient descent that has a momentum with dropout is one of the most popular training protocols. Based on that, more advanced methods (i.e., Maxout and Batch Normalization) have been proposed in recent years, but most stil...
متن کاملTowards Interpretable Deep Neural Networks by Leveraging Adversarial Examples
Deep neural networks (DNNs) have demonstrated impressive performance on a wide array of tasks, but they are usually considered opaque since internal structure and learned parameters are not interpretable. In this paper, we re-examine the internal representations of DNNs using adversarial images, which are generated by an ensembleoptimization algorithm. We find that: (1) the neurons in DNNs do n...
متن کاملFoveation-based Mechanisms Alleviate Adversarial Examples of Deep Neural Networks
Adversarial examples are images with visually imperceptible perturbations that 1 result in Deep Neural Networks (DNNs) fail. In this paper, we show that DNNs can 2 recover a similar level of accuracy prior to adding the adversarial perturbation, by 3 changing the image region in which the DNN is applied. This change of the input 4 image region is what we call foveation mechanism. There are many...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Mathematical statistics and learning
سال: 2023
ISSN: ['2520-2316', '2520-2324']
DOI: https://doi.org/10.4171/msl/41